翻訳と辞書
Words near each other
・ Backpackers (TV series)
・ Backpackers (web series)
・ BackpackersXpress
・ Backpacking (travel)
・ Backpacking (wilderness)
・ Backpacking Light Magazine
・ Backpacking with animals
・ Backpage
・ Backplane
・ Backplate and wing
・ Backporch Revolution
・ Backporting
・ Backpressure routing
・ Backpropagation
・ Backpropagation through structure
Backpropagation through time
・ Backrest
・ Backroad
・ Backroad (disambiguation)
・ Backroad Song
・ Backroads (album)
・ Backroads (film)
・ Backroads (song)
・ Backronym
・ Backs to the Land
・ Backsaw
・ Backscatter
・ Backscatter (disambiguation)
・ Backscatter (email)
・ Backscatter X-ray


Dictionary Lists
翻訳と辞書 辞書検索 [ 開発暫定版 ]
スポンサード リンク

Backpropagation through time : ウィキペディア英語版
Backpropagation through time

Backpropagation through time (BPTT) is a gradient-based technique for training certain types of recurrent neural networks. It can be used to train Elman networks. The algorithm was independently derived by numerous researchers
==Algorithm==

The training data for BPTT should be an ordered sequence of input-output pairs, \langle \mathbf_0,\mathbf_0 \rangle, \langle\mathbf_1,\mathbf_1 \rangle,\langle\mathbf_2,\mathbf_2\rangle,...,\langle\mathbf_,\mathbf_\rangle. An initial value must be specified for \mathbf_0. Typically, a vector of all zeros is used for this purpose.
BPTT begins by unfolding a recurrent neural network through time as shown in this figure. This recurrent neural network contains two feed-forward neural networks, ''f'' and ''g''. When the network is unfolded through time, the unfolded network contains ''k'' instances of ''f'' and one instance of ''g''. In the example shown, the network has been unfolded to a depth of ''k''=3.
Training then proceeds in a manner similar to training a feed-forward neural network with backpropagation, except that each epoch must run through the observations, \mathbf_t, in sequential order. Each training pattern consists of \langle\mathbf_t,\mathbf_t,\mathbf_,\mathbf_,...,\mathbf_,\mathbf_\rangle. (All of the actions for ''k'' time-steps are needed because the unfolded network contains inputs at each unfolded level.) Typically, backpropagation is applied in an online manner to update the weights as each training pattern is presented.
After each pattern is presented, and the weights have been updated, the weights in each instance of f (f_1, f_2, ..., f_k) are averaged together so that they all have the same weights. Also, \mathbf_ is calculated as \mathbf_=f(\mathbf_t, \mathbf_t), which provides the information necessary so that the algorithm can move on to the next time-step, ''t+1''.

抄文引用元・出典: フリー百科事典『 ウィキペディア(Wikipedia)
ウィキペディアで「Backpropagation through time」の詳細全文を読む



スポンサード リンク
翻訳と辞書 : 翻訳のためのインターネットリソース

Copyright(C) kotoba.ne.jp 1997-2016. All Rights Reserved.